452 research outputs found

    Systems of Linear Equations over F2\mathbb{F}_2 and Problems Parameterized Above Average

    Full text link
    In the problem Max Lin, we are given a system Az=bAz=b of mm linear equations with nn variables over F2\mathbb{F}_2 in which each equation is assigned a positive weight and we wish to find an assignment of values to the variables that maximizes the excess, which is the total weight of satisfied equations minus the total weight of falsified equations. Using an algebraic approach, we obtain a lower bound for the maximum excess. Max Lin Above Average (Max Lin AA) is a parameterized version of Max Lin introduced by Mahajan et al. (Proc. IWPEC'06 and J. Comput. Syst. Sci. 75, 2009). In Max Lin AA all weights are integral and we are to decide whether the maximum excess is at least kk, where kk is the parameter. It is not hard to see that we may assume that no two equations in Az=bAz=b have the same left-hand side and n=rankAn={\rm rank A}. Using our maximum excess results, we prove that, under these assumptions, Max Lin AA is fixed-parameter tractable for a wide special case: m2p(n)m\le 2^{p(n)} for an arbitrary fixed function p(n)=o(n)p(n)=o(n). Max rr-Lin AA is a special case of Max Lin AA, where each equation has at most rr variables. In Max Exact rr-SAT AA we are given a multiset of mm clauses on nn variables such that each clause has rr variables and asked whether there is a truth assignment to the nn variables that satisfies at least (12r)m+k2r(1-2^{-r})m + k2^{-r} clauses. Using our maximum excess results, we prove that for each fixed r2r\ge 2, Max rr-Lin AA and Max Exact rr-SAT AA can be solved in time 2O(klogk)+mO(1).2^{O(k \log k)}+m^{O(1)}. This improves 2O(k2)+mO(1)2^{O(k^2)}+m^{O(1)}-time algorithms for the two problems obtained by Gutin et al. (IWPEC 2009) and Alon et al. (SODA 2010), respectively

    A Hypergraph Dictatorship Test with Perfect Completeness

    Full text link
    A hypergraph dictatorship test is first introduced by Samorodnitsky and Trevisan and serves as a key component in their unique games based \PCP construction. Such a test has oracle access to a collection of functions and determines whether all the functions are the same dictatorship, or all their low degree influences are o(1).o(1). Their test makes q3q\geq3 queries and has amortized query complexity 1+O(logqq)1+O(\frac{\log q}{q}) but has an inherent loss of perfect completeness. In this paper we give an adaptive hypergraph dictatorship test that achieves both perfect completeness and amortized query complexity 1+O(logqq)1+O(\frac{\log q}{q}).Comment: Some minor correction

    When Can Limited Randomness Be Used in Repeated Games?

    Full text link
    The central result of classical game theory states that every finite normal form game has a Nash equilibrium, provided that players are allowed to use randomized (mixed) strategies. However, in practice, humans are known to be bad at generating random-like sequences, and true random bits may be unavailable. Even if the players have access to enough random bits for a single instance of the game their randomness might be insufficient if the game is played many times. In this work, we ask whether randomness is necessary for equilibria to exist in finitely repeated games. We show that for a large class of games containing arbitrary two-player zero-sum games, approximate Nash equilibria of the nn-stage repeated version of the game exist if and only if both players have Ω(n)\Omega(n) random bits. In contrast, we show that there exists a class of games for which no equilibrium exists in pure strategies, yet the nn-stage repeated version of the game has an exact Nash equilibrium in which each player uses only a constant number of random bits. When the players are assumed to be computationally bounded, if cryptographic pseudorandom generators (or, equivalently, one-way functions) exist, then the players can base their strategies on "random-like" sequences derived from only a small number of truly random bits. We show that, in contrast, in repeated two-player zero-sum games, if pseudorandom generators \emph{do not} exist, then Ω(n)\Omega(n) random bits remain necessary for equilibria to exist

    A shortcut to (sun)flowers: Kernels in logarithmic space or linear time

    Full text link
    We investigate whether kernelization results can be obtained if we restrict kernelization algorithms to run in logarithmic space. This restriction for kernelization is motivated by the question of what results are attainable for preprocessing via simple and/or local reduction rules. We find kernelizations for d-Hitting Set(k), d-Set Packing(k), Edge Dominating Set(k) and a number of hitting and packing problems in graphs, each running in logspace. Additionally, we return to the question of linear-time kernelization. For d-Hitting Set(k) a linear-time kernelization was given by van Bevern [Algorithmica (2014)]. We give a simpler procedure and save a large constant factor in the size bound. Furthermore, we show that we can obtain a linear-time kernel for d-Set Packing(k) as well.Comment: 18 page

    Parallel Repetition of Entangled Games with Exponential Decay via the Superposed Information Cost

    Get PDF
    In a two-player game, two cooperating but non communicating players, Alice and Bob, receive inputs taken from a probability distribution. Each of them produces an output and they win the game if they satisfy some predicate on their inputs/outputs. The entangled value ω(G)\omega^*(G) of a game GG is the maximum probability that Alice and Bob can win the game if they are allowed to share an entangled state prior to receiving their inputs. The nn-fold parallel repetition GnG^n of GG consists of nn instances of GG where the players receive all the inputs at the same time and produce all the outputs at the same time. They win GnG^n if they win each instance of GG. In this paper we show that for any game GG such that ω(G)=1ε<1\omega^*(G) = 1 - \varepsilon < 1, ω(Gn)\omega^*(G^n) decreases exponentially in nn. First, for any game GG on the uniform distribution, we show that ω(Gn)=(1ε2)Ω(nlog(IO)log(ε))\omega^*(G^n) = (1 - \varepsilon^2)^{\Omega\left(\frac{n}{\log(|I||O|)} - |\log(\varepsilon)|\right)}, where I|I| and O|O| are the sizes of the input and output sets. From this result, we show that for any entangled game GG, ω(Gn)(1ε2)Ω(nQlog(IO)log(ε)Q)\omega^*(G^n) \le (1 - \varepsilon^2)^{\Omega(\frac{n}{Q\log(|I||O|)} - \frac{|\log(\varepsilon)|}{Q})} where pp is the input distribution of GG and Q=I2maxxypxy2minxypxyQ= \frac{|I|^2 \max_{xy} p_{xy}^2 }{\min_{xy} p_{xy} }. This implies parallel repetition with exponential decay as long as minxy{pxy}0\min_{xy} \{p_{xy}\} \neq 0 for general games. To prove this parallel repetition, we introduce the concept of \emph{Superposed Information Cost} for entangled games which is inspired from the information cost used in communication complexity.Comment: In the first version of this paper we presented a different, stronger Corollary 1 but due to an error in the proof we had to modify it in the second version. This third version is a minor update. We correct some typos and re-introduce a proof accidentally commented out in the second versio

    On the Complexity of Computing Two Nonlinearity Measures

    Full text link
    We study the computational complexity of two Boolean nonlinearity measures: the nonlinearity and the multiplicative complexity. We show that if one-way functions exist, no algorithm can compute the multiplicative complexity in time 2O(n)2^{O(n)} given the truth table of length 2n2^n, in fact under the same assumption it is impossible to approximate the multiplicative complexity within a factor of (2ϵ)n/2(2-\epsilon)^{n/2}. When given a circuit, the problem of determining the multiplicative complexity is in the second level of the polynomial hierarchy. For nonlinearity, we show that it is #P hard to compute given a function represented by a circuit

    On the Approximability of Digraph Ordering

    Full text link
    Given an n-vertex digraph D = (V, A) the Max-k-Ordering problem is to compute a labeling :V[k]\ell : V \to [k] maximizing the number of forward edges, i.e. edges (u,v) such that \ell(u) < \ell(v). For different values of k, this reduces to Maximum Acyclic Subgraph (k=n), and Max-Dicut (k=2). This work studies the approximability of Max-k-Ordering and its generalizations, motivated by their applications to job scheduling with soft precedence constraints. We give an LP rounding based 2-approximation algorithm for Max-k-Ordering for any k={2,..., n}, improving on the known 2k/(k-1)-approximation obtained via random assignment. The tightness of this rounding is shown by proving that for any k={2,..., n} and constant ε>0\varepsilon > 0, Max-k-Ordering has an LP integrality gap of 2 - ε\varepsilon for nΩ(1/loglogk)n^{\Omega\left(1/\log\log k\right)} rounds of the Sherali-Adams hierarchy. A further generalization of Max-k-Ordering is the restricted maximum acyclic subgraph problem or RMAS, where each vertex v has a finite set of allowable labels SvZ+S_v \subseteq \mathbb{Z}^+. We prove an LP rounding based 42/(2+1)2.3444\sqrt{2}/(\sqrt{2}+1) \approx 2.344 approximation for it, improving on the 222.8282\sqrt{2} \approx 2.828 approximation recently given by Grandoni et al. (Information Processing Letters, Vol. 115(2), Pages 182-185, 2015). In fact, our approximation algorithm also works for a general version where the objective counts the edges which go forward by at least a positive offset specific to each edge. The minimization formulation of digraph ordering is DAG edge deletion or DED(k), which requires deleting the minimum number of edges from an n-vertex directed acyclic graph (DAG) to remove all paths of length k. We show that both, the LP relaxation and a local ratio approach for DED(k) yield k-approximation for any k[n]k\in [n].Comment: 21 pages, Conference version to appear in ESA 201

    Scheduling over Scenarios on Two Machines

    Get PDF
    We consider scheduling problems over scenarios where the goal is to find a single assignment of the jobs to the machines which performs well over all possible scenarios. Each scenario is a subset of jobs that must be executed in that scenario and all scenarios are given explicitly. The two objectives that we consider are minimizing the maximum makespan over all scenarios and minimizing the sum of the makespans of all scenarios. For both versions, we give several approximation algorithms and lower bounds on their approximability. With this research into optimization problems over scenarios, we have opened a new and rich field of interesting problems.Comment: To appear in COCOON 2014. The final publication is available at link.springer.co

    On the NP-Hardness of Approximating Ordering Constraint Satisfaction Problems

    Full text link
    We show improved NP-hardness of approximating Ordering Constraint Satisfaction Problems (OCSPs). For the two most well-studied OCSPs, Maximum Acyclic Subgraph and Maximum Betweenness, we prove inapproximability of 14/15+ϵ14/15+\epsilon and 1/2+ϵ1/2+\epsilon. An OCSP is said to be approximation resistant if it is hard to approximate better than taking a uniformly random ordering. We prove that the Maximum Non-Betweenness Problem is approximation resistant and that there are width-mm approximation-resistant OCSPs accepting only a fraction 1/(m/2)!1 / (m/2)! of assignments. These results provide the first examples of approximation-resistant OCSPs subject only to P \neq \NP

    Constant-degree graph expansions that preserve the treewidth

    Full text link
    Many hard algorithmic problems dealing with graphs, circuits, formulas and constraints admit polynomial-time upper bounds if the underlying graph has small treewidth. The same problems often encourage reducing the maximal degree of vertices to simplify theoretical arguments or address practical concerns. Such degree reduction can be performed through a sequence of splittings of vertices, resulting in an _expansion_ of the original graph. We observe that the treewidth of a graph may increase dramatically if the splittings are not performed carefully. In this context we address the following natural question: is it possible to reduce the maximum degree to a constant without substantially increasing the treewidth? Our work answers the above question affirmatively. We prove that any simple undirected graph G=(V, E) admits an expansion G'=(V', E') with the maximum degree <= 3 and treewidth(G') <= treewidth(G)+1. Furthermore, such an expansion will have no more than 2|E|+|V| vertices and 3|E| edges; it can be computed efficiently from a tree-decomposition of G. We also construct a family of examples for which the increase by 1 in treewidth cannot be avoided.Comment: 12 pages, 6 figures, the main result used by quant-ph/051107
    corecore